Efficient Active Learning of Halfspaces: an Aggressive Approach
نویسندگان
چکیده
We study pool-based active learning of halfspaces. We revisit the aggressive approach for active learning in the realizable case, and show that it can be made efficient and practical, while also having theoretical guarantees under reasonable assumptions. We further show, both theoretically and experimentally, that it can be preferable to mellow approaches. Our efficient aggressive active learner of half-spaces has formal approximation guarantees that hold when the pool is separable with a margin. While our analysis is focused on the realizable setting, we show that a simple heuristic allows using the same algorithm successfully for pools with low error as well. We further compare the aggressive approach to the mellow approach, and prove that there are cases in which the aggressive approach results in significantly better label complexity compared to the mellow approach. Experiments demonstrate that substantial improvements in label complexity can be achieved using the aggressive approach, in realizable and low-error settings.
منابع مشابه
Active Learning of Halfspaces
We study pool-based active learning of half-spaces. In this setting a learner receives a pool of unlabeled examples, and can iteratively query a teacher for the labels of examples from the pool. The goal of the learner is to return a low-error prediction rule for the labels of the examples, using a small number of queries. Most active learning approaches can be loosely described as more ‘aggres...
متن کاملActive Learning of Halfspaces under a Margin Assumption
We derive and analyze a new, efficient, pool-based active learning algorithm for halfspaces, called ALuMA. Most previous algorithms show exponential improvement in the label complexity assuming that the distribution over the instance space is close to uniform. This assumption rarely holds in practical applications. Instead, we study the label complexity under a large-margin assumption—a much mo...
متن کاملConvex Optimization for Active Learning with Large Margins
In this paper we show how large margin assumptions make it possible to use ideas and algorithms from convex optimization for active learning. This provides an alternative and complementary approach to standard algorithms for active learning. These algorithms appear to be robust and provide approximately correct hypotheses with probability one, as opposed to the standard PAC learning results. In...
متن کاملEfficient Active Learning of Halfspaces via Query Synthesis
Active learning is a subfield of machine learning that has been successfully used in many applications including text classification and bioinformatics. One of the fundamental branches of active learning is query synthesis, where the learning agent constructs artificial queries from scratch in order to reveal sensitive information about the true decision boundary. Nevertheless, the existing lit...
متن کاملOpen Problem: The Statistical Query Complexity of Learning Sparse Halfspaces
We consider the long-open problem of attribute-efficient learning of halfspaces. In this problem the learner is given random examples labeled by an unknown halfspace function f on R. Further f is r-sparse, that is it depends on at most r out of n variables. An attribute-efficient learning algorithm is an algorithm that can output a hypothesis close to f using a polynomial in r and log n number ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2013